washington university
Beyond Chains: Bridging Large Language Models and Knowledge Bases in Complex Question Answering
Zhu, Yihua, Liu, Qianying, Aizawa, Akiko, Shimodaira, Hidetoshi
Knowledge Base Question Answering (KBQA) aims to answer natural language questions using structured knowledge from KBs. While LLM-only approaches offer generalization, they suffer from outdated knowledge, hallucinations, and lack of transparency. Chain-based KG-RAG methods address these issues by incorporating external KBs, but are limited to simple chain-structured questions due to the absence of planning and logical structuring. Inspired by semantic parsing methods, we propose PDRR: a four-stage framework consisting of Predict, Decompose, Retrieve, and Reason. Our method first predicts the question type and decomposes the question into structured triples. Then retrieves relevant information from KBs and guides the LLM as an agent to reason over and complete the decomposed triples. Experimental results demonstrate that PDRR consistently outperforms existing methods across various LLM backbones and achieves superior performance on both chain-structured and non-chain complex questions.
- Europe > France (0.05)
- Europe > Netherlands > Gelderland > Nijmegen (0.05)
- North America > United States > Tennessee (0.05)
- (23 more...)
- Leisure & Entertainment (1.00)
- Media > Music (0.49)
Multidimensional Bayesian Active Machine Learning of Working Memory Task Performance
Marticorena, Dom CP, Wissmann, Chris, Lu, Zeyu, Barbour, Dennis L
While adaptive experimental design has outgrown one-dimensional, staircase-based adaptations, most cognitive experiments still control a single factor and summarize performance with a scalar. We show a validation of a Bayesian, two-axis, active-classification approach, carried out in an immersive virtual testing environment for a 5-by-5 working-memory reconstruction task. Two variables are controlled: spatial load L (number of occupied tiles) and feature-binding load K (number of distinct colors) of items. Stimulus acquisition is guided by posterior uncertainty of a nonparametric Gaussian Process (GP) probabilistic classifier, which outputs a surface over (L, K) rather than a single threshold or max span value. In a young adult population, we compare GP-driven Adaptive Mode (AM) with a traditional adaptive staircase Classic Mode (CM), which varies L only at K = 3. Parity between the methods is achieved for this cohort, with an intraclass coefficient of 0.755 at K = 3. Additionally, AM reveals individual differences in interactions between spatial load and feature binding. AM estimates converge more quickly than other sampling strategies, demonstrating that only about 30 samples are required for accurate fitting of the full model.
- North America > United States > Missouri > St. Louis County > St. Louis (0.04)
- Europe > Ukraine (0.04)
- Europe > Finland > Uusimaa > Helsinki (0.04)
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (0.68)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models (0.46)
GraphSeqLM: A Unified Graph Language Framework for Omic Graph Learning
Zhang, Heming, Huang, Di, Chen, Yixin, Li, Fuhai
The integration of multi-omic data is pivotal for understanding complex diseases, but its high dimensionality and noise present significant challenges. Graph Neural Networks (GNNs) offer a robust framework for analyzing large-scale signaling pathways and protein-protein interaction networks, yet they face limitations in expressivity when capturing intricate biological relationships. To address this, we propose Graph Sequence Language Model (GraphSeqLM), a framework that enhances GNNs with biological sequence embeddings generated by Large Language Models (LLMs). These embeddings encode structural and biological properties of DNA, RNA, and proteins, augmenting GNNs with enriched features for analyzing sample-specific multi-omic data. By integrating topological, sequence-derived, and biological information, GraphSeqLM demonstrates superior predictive accuracy and outperforms existing methods, paving the way for more effective multi-omic data integration in precision medicine.
- Oceania > Australia > New South Wales > Sydney (0.05)
- North America > United States > Missouri > St. Louis County > St. Louis (0.05)
- Asia > Japan > Honshū > Kansai > Kyoto Prefecture > Kyoto (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Health & Medicine > Pharmaceuticals & Biotechnology (1.00)
Improving Normative Modeling for Multi-modal Neuroimaging Data using mixture-of-product-of-experts variational autoencoders
Kumar, Sayantan, Payne, Philip, Sotiras, Aristeidis
Normative models in neuroimaging learn the brain patterns of healthy population distribution and estimate how disease subjects like Alzheimer's Disease (AD) deviate from the norm. Existing variational autoencoder (VAE)-based normative models using multimodal neuroimaging data aggregate information from multiple modalities by estimating product or averaging of unimodal latent posteriors. This can often lead to uninformative joint latent distributions which affects the estimation of subject-level deviations. In this work, we addressed the prior limitations by adopting the Mixture-of-Product-of-Experts (MoPoE) technique which allows better modelling of the joint latent posterior. Our model labelled subjects as outliers by calculating deviations from the multimodal latent space. Further, we identified which latent dimensions and brain regions were associated with abnormal deviations due to AD pathology.
- North America > United States > Montana (0.04)
- North America > United States > Massachusetts > Suffolk County > Boston (0.04)
- Health & Medicine > Health Care Technology (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
- Health & Medicine > Therapeutic Area > Neurology > Alzheimer's Disease (0.92)
AI may have an 'eye' on growing babies: Could predict premature birth as early as 31 weeks
Fox News medical contributor Dr. Marc Siegel joins'Fox & Friends' to discuss the benefits of artificial intelligence in the medical industry if used with caution. About 10% of all infants born in the U.S. in 2021 were preterm -- which means they were delivered earlier than 37 weeks of pregnancy, per the Centers for Disease Control and Prevention (CDC). Preterm births also make up about 16% of infant deaths. Now, researchers from Washington University in St. Louis, Missouri, are looking to improve those odds through the use of artificial intelligence. They developed a deep learning model that can predict preterm births by analyzing electrical activity in the woman's uterus during pregnancy -- then they tested the model in a study that was published in the medical journal PLOS One.
- North America > United States > Missouri > St. Louis County > St. Louis (0.25)
- North America > United States > Colorado > Denver County > Denver (0.05)
- Health & Medicine > Therapeutic Area > Obstetrics/Gynecology (1.00)
- Health & Medicine > Public Health > Maternal Health (1.00)
Global Performance Disparities Between English-Language Accents in Automatic Speech Recognition
DiChristofano, Alex, Shuster, Henry, Chandra, Shefali, Patwari, Neal
However, many users are familiar with the frustrating experience of repeatedly not being understood by their voice assistant [16], so much so that frustration with ASR has become a culturally-shared source of comedy [4, 32]. Bias auditing of ASR services has quantified these experiences. English language ASR has higher error rates: for Black Americans compared to white Americans [24, 45], for stigmatised British accents compared to favored British accents [28], for Scottish speakers compared to speakers from California and New Zealand [44], for speakers whose first language is a tone language compared to those whose first language is not [2], for speakers with Indian accents compared to speakers who with "American" accents [31], for speakers whose first language is English compared to those for whom it is not [28]. It should go without saying, but everyone has an accent - there is no "unaccented" version of English [26]. Due to colonization and globalization, different Englishes are spoken around the world. While some English accents may be favored by those with class, race, and national origin privilege [28], there is no technical barrier to building an ASR system which works well on any particular accent. So we are left with the question, why does ASR performance vary as it does as a function of the global English accent spoken?
- Oceania > New Zealand (0.24)
- North America > United States > Missouri > St. Louis County > St. Louis (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (25 more...)
- Health & Medicine (1.00)
- Education (0.93)
- Government > Regional Government (0.46)
Predicting Intraoperative Hypoxemia with Hybrid Inference Sequence Autoencoder Networks
Liu, Hanyang, Montana, Michael C., Li, Dingwen, Renfroe, Chase, Kannampallil, Thomas, Lu, Chenyang
We present an end-to-end model using streaming physiological time series to predict near-term risk for hypoxemia, a rare, but life-threatening condition known to cause serious patient harm during surgery. Inspired by the fact that a hypoxemia event is defined based on a future sequence of low SpO2 (i.e., blood oxygen saturation) instances, we propose the hybrid inference network (hiNet) that makes hybrid inference on both future low SpO2 instances and hypoxemia outcomes. hiNet integrates 1) a joint sequence autoencoder that simultaneously optimizes a discriminative decoder for label prediction, and 2) two auxiliary decoders trained for data reconstruction and forecast, which seamlessly learn contextual latent representations that capture the transition from present states to future states. All decoders share a memory-based encoder that helps capture the global dynamics of patient measurement. For a large surgical cohort of 72,081 surgeries at a major academic medical center, our model outperforms strong baselines including the model used by the state-of-the-art hypoxemia prediction system. With its capability to make real-time predictions of near-term hypoxemic at clinically acceptable alarm rates, hiNet shows promise in improving clinical decision making and easing burden of perioperative care.
- North America > United States > Georgia > Fulton County > Atlanta (0.05)
- North America > United States > Missouri > St. Louis County > St. Louis (0.05)
- North America > United States > Montana (0.04)
GitHub - jeffheaton/t81_558_deep_learning: Washington University (in St. Louis) Course T81-558: Applications of Deep Neural Networks
The content of this course changes as technology evolves, to keep up to date with changes follow me on GitHub. Deep learning is a group of exciting new technologies for neural networks. Through a combination of advanced training techniques and neural network architectural components, it is now possible to create neural networks that can handle tabular data, images, text, and audio as both input and output. Deep learning allows a neural network to learn hierarchies of information in a way that is like the function of the human brain. This course will introduce the student to classic neural network structures, Convolution Neural Networks (CNN), Long Short-Term Memory (LSTM), Gated Recurrent Neural Networks (GRU), General Adversarial Networks (GAN) and reinforcement learning.